On Parameter Tying by Quantization
نویسندگان
چکیده
The maximum likelihood estimator (MLE) is generally asymptotically consistent but is susceptible to overfitting. To combat this problem, regularization methods which reduce the variance at the cost of (slightly) increasing the bias are often employed in practice. In this paper, we present an alternative variance reduction (regularization) technique that quantizes the MLE estimates as a post processing step, yielding a smoother model having several tied parameters. We provide and prove error bounds for our new technique and demonstrate experimentally that it often yields models having higher test-set log-likelihood than the ones learned using the MLE. We also propose a new importance sampling algorithm for fast approximate inference in models having several tied parameters. Our experiments show that our new inference algorithm is superior to existing approaches such as Gibbs sampling and MCSAT on models having tied parameters, learned using our quantization-based approach.
منابع مشابه
Parameter tying for flexible speech recognition
This paper presents two parameter tying techniques which enable a trade-off between computational cost and recognition performances of a speaker independent flexible speech recognition system working over the telephone network. Parameter tying is conducted at phonetic and acoustic levels. At the phonetic level, allophone and triphone based phonetic modeling are used simultaneously to achieve th...
متن کاملDeep Learning: Autodiff, Parameter Tying and Backprop Through Time∗
How to do parameter tying and how this relates to Backprop through time.
متن کاملSubstate tying with combined parameter training and reduction in tied-mixture HMM design
Two approaches are proposed for the design of tied-mixture hidden Markov models (TMHMM). One approach improves parameter sharing via partial tying of TMHMM states. To facilitate tying at the substate level, the state emission probabilities are constructed in two stages or, equivalently, are viewed as a “mixture of mixtures of Gaussians.” This paradigm allows, and is complemented with, an optimi...
متن کاملFlexible Parameter Tying for Conversational Speech Recognition
Modeling pronunciation variation is key for recognizing conversational speech. Previous efforts on pronunciation modeling by modifying dictionaries only yielded marginal improvement. Due to complex interaction between dictionaries and acoustic models, we believe a pronunciation modeling scheme is plausible only when closely coupled with the underlying acoustic model. This paper explores the use...
متن کاملSharing Trend Information of Trajecto
In this paper, the reduction method of number of parameters in the segmental-feature HMM (SFHMM) can be considered. It is reported that the SFHMM shows better results than conventional HMM in the previous studies. However, its number of parameters is greater than that of HMM. Therefore, there is a need for new approach that reduces the number of parameters. The trajectories are used for the aco...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2016